Understanding Implicit Regularization in Over-Parameterized Single Index Model

نویسندگان

چکیده

In this paper, we leverage over-parameterization to design regularization-free algorithms for the high-dimensional single index model and provide theoretical guarantees induced implicit regularization phenomenon. Specifically, study both vector matrix models where link function is nonlinear unknown, signal parameter either a sparse or low-rank symmetric matrix, response variable can be heavy-tailed. To gain better understanding of role played by without excess technicality, assume that distribution covariates known priori. For settings, construct an over-parameterized least-squares loss employing score transform robust truncation step designed specifically heavy-tailed data. We propose estimate true applying gradient descent function. When initialization close origin stepsize sufficiently small, prove obtained solution achieves minimax optimal statistical rates convergence in cases. addition, our experimental results support findings also demonstrate methods empirically outperform classical with explicit terms $\ell_2$-statistical rate selection consistency.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Identifiability of Model Properties in Over-Parameterized Model Classes

Classical learning theory is based on a tight linkage between hypothesis space (a class of function on a domain X), data space (function-value examples (x, f(x))), and the space of queries for the learned model (predicting function values for new examples x). However, in many learning scenarios the 3-way association between hypotheses, data, and queries can really be much looser. Model classes ...

متن کامل

Implicit Regularization in Deep Learning

In an attempt to better understand generalization in deep learning, we study several possible explanations. We show that implicit regularization induced by the optimization method is playing a key role in generalization and success of deep learning models. Motivated by this view, we study how different complexity measures can ensure generalization and explain how optimization algorithms can imp...

متن کامل

Implicit Regularization in Matrix Factorization

We study implicit regularization when optimizing an underdetermined quadratic objective over a matrix X with gradient descent on a factorization of X . We conjecture and provide empirical and theoretical evidence that with small enough step sizes and initialization close enough to the origin, gradient descent on a full dimensional factorization converges to the minimum nuclear norm solution.

متن کامل

Algorithmic Regularization in Over-parameterized Matrix Sensing and Neural Networks with Quadratic Activations

We show that the gradient descent algorithm provides an implicit regularization effect in the learning of over-parameterizedmatrix factorization models and one-hidden-layer neural networks with quadratic activations. Concretely, we show that given Õ(dr) random linear measurements of a rank r positive semidefinite matrix X, we can recover X by parameterizing it by UU with U ∈ R and minimizing th...

متن کامل

Sparse single-index model

Let (X, Y ) be a random pair taking values in Rp × R. In the socalled single-index model, one has Y = f?(θ?TX) + W , where f? is an unknown univariate measurable function, θ? is an unknown vector in Rd, and W denotes a random noise satisfying E[W |X] = 0. The single-index model is known to offer a flexible way to model a variety of high-dimensional real-world phenomena. However, despite its rel...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of the American Statistical Association

سال: 2022

ISSN: ['0162-1459', '1537-274X', '2326-6228', '1522-5445']

DOI: https://doi.org/10.1080/01621459.2022.2044824